Kullback–Leibler divergence
This captures the difference between the entropy of and the “wrong” entropy of measured with ().
A common way to measure “distance” between two distributions. It’s not a real distance because it’s not symmetric. The symmetrized version is called Jensen–Shannon divergence.
See also Gibbs’ inequality